What are some abliterated models to run in Ollama on a an Intel Core i7-6700 CPU @ 3.40GHz with 16G of RAM and a GeForce GTX 745?

It's fantastic you're exploring Llama models! Let's talk about which ones would be realistic to run on your setup:

Understanding the Constraints

Ollama Model Considerations

Recommended Models (Based on Size)

Here are a few model sizes commonly considered suitable for your setup:

  1. 7B Parameter Model (Smaller):
  2. Potential Challenges: May experience limitations in accuracy and context handling.
  3. Suitable for: Experimentation, basic text generation, Q&A.
  4. 65B Parameter Model (Larger):
  5. Challenges: Requires significant GPU memory and might be too large for your machine.
  6. Best for: Research, advanced language tasks that demand more computational power but are very computationally demanding.

Tips for Success with Ollama

  1. Experimentation: Start small! Test different model sizes to see how they perform on your hardware.
  2. Quantization is Key: Use tools like transformers' quantization techniques for memory efficiency.
  3. Resource Monitoring: Pay attention to CPU and GPU usage while running Ollama. You may need to adjust parameters if needed.

Important Notes:

Let me know if you have other questions or want help setting up the necessary tools!

Back to the List